Regularized singular value decomposition: a sparse dimension reduction technique
نویسنده
چکیده
Singular value decomposition (SVD) is a useful multivariate technique for dimension reduction. It has been successfully applied to analyze microarray data, where the eigen vectors are called eigen-genes/arrays. One weakness associated with the SVD is the interpretation. The eigen-genes are essentially linear combinations of all the genes. It is desirable to have sparse SVD, which retains the dimension reduction property but also the eigen vectors are linear combinations of only a small subset of genes. In this paper we formally propose a statistical framework for sparse SVD, which is a generalization of traditional SVD. Our formulation of sparse SVD also bears the close connection to the penalized t-/F-statistics for differential gene expression detection, which can be viewed as the supervised sparse dimension reduction technique.
منابع مشابه
Sparse Principal Component Analysis via Regularized Low Rank Matrix Approximation
Principal component analysis (PCA) is a widely used tool for data analysis and dimension reduction in applications throughout science and engineering. However, the principal components (PCs) can sometimes be difficult to interpret, because they are linear combinations of all the original variables. To facilitate interpretation, sparse PCA produces modified PCs with sparse loadings, i.e. loading...
متن کاملAnalysis of Singular Value Decomposition as a Dimensionality Reduction Technique and Sparse Representation Classifier as a Post Classifier for the Classification of Epilepsy Risk Levels from EEG Signals
The main aim of this paper is to perform the analysis of Singular Value Decomposition (SVD) as a Dimensionality Reduction technique and Sparse Representation Classifier (SRC) as a Post Classifier for the Classification of Epilepsy Risk levels from Electroencephalography signals. The data acquisition of the EEG signals is performed initially. Then SVD is applied here as a dimensionality reductio...
متن کاملLarge-scale Inversion of Magnetic Data Using Golub-Kahan Bidiagonalization with Truncated Generalized Cross Validation for Regularization Parameter Estimation
In this paper a fast method for large-scale sparse inversion of magnetic data is considered. The L1-norm stabilizer is used to generate models with sharp and distinct interfaces. To deal with the non-linearity introduced by the L1-norm, a model-space iteratively reweighted least squares algorithm is used. The original model matrix is factorized using the Golub-Kahan bidiagonalization that proje...
متن کاملFast Linear Discriminant Analysis using QR Decomposition and Regularization
Linear Discriminant Analysis (LDA) is among the most optimal dimension reduction methods for classification, which provides a high degree of class separability for numerous applications from science and engineering. However, problems arise with this classical method when one or both of the scatter matrices is singular. Singular scatter matrices are not unusual in many applications, especially f...
متن کاملSparse Principal Component Analysis Incorporating Stability Selection
Principal component analysis (PCA) is a popular dimension reduction method that approximates a numerical data matrix by seeking principal components (PC), i.e. linear combinations of variables that captures maximal variance. Since each PC is a linear combination of all variables of a data set, interpretation of the PCs can be difficult, especially in high-dimensional data. In order to find ’spa...
متن کامل